recognition algorithm
Interpretable Recognition of Cognitive Distortions in Natural Language Texts
Kolonin, Anton, Arinicheva, Anna
We propose a new approach to multi-factor classification of natural language texts based on weighted structured patterns such as N-grams, taking into account the heterarchical relationships between them, applied to solve such a socially impactful problem as the automation of detection of specific cognitive distortions in psychological care, relying on an interpretable, robust and transparent artificial intelligence model. The proposed recognition and learning algorithms improve the current state of the art in this field. The improvement is tested on two publicly available datasets, with significant improvements over literature-known F1 scores for the task, with optimal hyper-parameters determined, having code and models available for future use by the community.
- Europe > Switzerland (0.04)
- North America > United States > Florida > Miami-Dade County > Miami (0.04)
- North America > United States > Connecticut > New Haven County > Madison (0.04)
- (5 more...)
DIJE: Dense Image Jacobian Estimation for Robust Robotic Self-Recognition and Visual Servoing
Toshimitsu, Yasunori, Kawaharazuka, Kento, Miki, Akihiro, Okada, Kei, Inaba, Masayuki
For robots to move in the real world, they must first correctly understand the state of its own body and the tools that it holds. In this research, we propose DIJE, an algorithm to estimate the image Jacobian for every pixel. It is based on an optical flow calculation and a simplified Kalman Filter that can be efficiently run on the whole image in real time. It does not rely on markers nor knowledge of the robotic structure. We use the DIJE in a self-recognition process which can robustly distinguish between movement by the robot and by external entities, even when the motion overlaps. We also propose a visual servoing controller based on DIJE, which can learn to control the robot's body to conduct reaching movements or bimanual tool-tip control. The proposed algorithms were implemented on a physical musculoskeletal robot and its performance was verified. We believe that such global estimation of the visuomotor policy has the potential to be extended into a more general framework for manipulation.
Machine vision-aware quality metrics for compressed image and video assessment
Dremin, Mikhail, Kozhemyakov, Konstantin, Molodetskikh, Ivan, Kirill, Malakhov, Sagitov, Artur, Vatolin, Dmitriy
A main goal in developing video-compression algorithms is to enhance human-perceived visual quality while maintaining file size. But modern video-analysis efforts such as detection and recognition, which are integral to video surveillance and autonomous vehicles, involve so much data that they necessitate machine-vision processing with minimal human intervention. In such cases, the video codec must be optimized for machine vision. This paper explores the effects of compression on detection and recognition algorithms (objects, faces, and license plates) and introduces novel full-reference image/video-quality metrics for each task, tailored to machine vision. Experimental results indicate our proposed metrics correlate better with the machine-vision results for the respective tasks than do existing image/video-quality metrics.
- Europe > Switzerland > Zürich > Zürich (0.14)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
Object Recognition in Human Computer Interaction:- A Comparative Analysis
Ranade, Kaushik, Khule, Tanmay, More, Riddhi
Human-computer interaction (HCI) has been a widely researched area for many years, with continuous advancements in technology leading to the development of new techniques that change the way we interact with computers. With the recent advent of powerful computers, we recognize human actions and interact accordingly, thus revolutionizing the way we interact with computers. The purpose of this paper is to provide a comparative analysis of various algorithms used for recognizing user faces and gestures in the context of computer vision and HCI. This study aims to explore and evaluate the performance of different algorithms in terms of accuracy, robustness, and efficiency. This study aims to provide a comprehensive analysis of algorithms for face and gesture recognition in the context of computer vision and HCI, with the goal of improving the design and development of interactive systems that are more intuitive, efficient, and user-friendly.
- Information Technology > Human Computer Interaction (1.00)
- Information Technology > Artificial Intelligence > Vision > Gesture Recognition (1.00)
- Information Technology > Artificial Intelligence > Vision > Face Recognition (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Selective, Interpretable, and Motion Consistent Privacy Attribute Obfuscation for Action Recognition
Ilic, Filip, Zhao, He, Pock, Thomas, Wildes, Richard P.
Concerns for the privacy of individuals captured in public imagery have led to privacy-preserving action recognition. Existing approaches often suffer from issues arising through obfuscation being applied globally and a lack of interpretability. Global obfuscation hides privacy sensitive regions, but also contextual regions important for action recognition. Lack of interpretability erodes trust in these new technologies. We highlight the limitations of current paradigms and propose a solution: Human selected privacy templates that yield interpretability by design, an obfuscation scheme that selectively hides attributes and also induces temporal consistency, which is important in action recognition. Our approach is architecture agnostic and directly modifies input imagery, while existing approaches generally require architecture training. Our approach offers more flexibility, as no retraining is required, and outperforms alternatives on three widely used datasets.
Unsupervised Template Learning for Fine-Grained Object Recognition
Fine-grained recognition refers to a subordinate level of recognition, such as recognizing different species of animals and plants. It differs from recognition of basic categories, such as humans, tables, and computers, in that there are global similarities in shape and structure shared cross different categories, and the differences are in the details of object parts. We suggest that the key to identifying the fine-grained differences lies in finding the right alignment of image regions that contain the same object parts. We propose a template model for the purpose, which captures common shape patterns of object parts, as well as the cooccurrence relation of the shape patterns. Once the image regions are aligned, extracted features are used for classification. Learning of the template model is efficient, and the recognition results we achieve significantly outperform the stateof-the-art algorithms.
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.47)
Facial recognition technology and human raters can predict political orientation from images of expressionless faces even when controlling for demographics and self-presentation
Kosinski, Michal, Khambatta, Poruz, Wang, Yilun
Carefully standardized facial images of 591 participants were taken in the laboratory, while controlling for self-presentation, facial expression, head orientation, and image properties. They were presented to human raters and a facial recognition algorithm: both humans (r=.21) and the algorithm (r=.22) could predict participants' scores on a political orientation scale (Cronbach's alpha=.94) decorrelated with age, gender, and ethnicity. These effects are on par with how well job interviews predict job success, or alcohol drives aggressiveness. Algorithm's predictive accuracy was even higher (r=.31) when it leveraged information on participants' age, gender, and ethnicity. Moreover, the associations between facial appearance and political orientation seem to generalize beyond our sample: The predictive model derived from standardized images (while controlling for age, gender, and ethnicity) could predict political orientation (r=.13) from naturalistic images of 3,401 politicians from the U.S., UK, and Canada. The analysis of facial features associated with political orientation revealed that conservatives tended to have larger lower faces. The predictability of political orientation from standardized images has critical implications for privacy, the regulation of facial recognition technology, and understanding the origins and consequences of political orientation.
- North America > Canada (0.25)
- North America > United States > Mississippi (0.04)
- North America > United States > Colorado > Boulder County > Boulder (0.04)
- (2 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- Government > Regional Government (0.93)
- Health & Medicine (0.68)
- Government > Voting & Elections (0.68)
- Information Technology (0.67)
Vision- and tactile-based continuous multimodal intention and attention recognition for safer physical human-robot interaction
Wong, Christopher Yee, Vergez, Lucas, Suleiman, Wael
Employing skin-like tactile sensors on robots enhances both the safety and usability of collaborative robots by adding the capability to detect human contact. Unfortunately, simple binary tactile sensors alone cannot determine the context of the human contact -- whether it is a deliberate interaction or an unintended collision that requires safety manoeuvres. Many published methods classify discrete interactions using more advanced tactile sensors or by analysing joint torques. Instead, we propose to augment the intention recognition capabilities of simple binary tactile sensors by adding a robot-mounted camera for human posture analysis. Different interaction characteristics, including touch location, human pose, and gaze direction, are used to train a supervised machine learning algorithm to classify whether a touch is intentional or not with an F1-score of 86%. We demonstrate that multimodal intention recognition is significantly more accurate than monomodal analyses with the collaborative robot Baxter. Furthermore, our method can also continuously monitor interactions that fluidly change between intentional or unintentional by gauging the user's attention through gaze. If a user stops paying attention mid-task, the proposed intention and attention recognition algorithm can activate safety features to prevent unsafe interactions. We also employ a feature reduction technique that reduces the number of inputs to five to achieve a more generalized low-dimensional classifier. This simplification both reduces the amount of training data required and improves real-world classification accuracy. It also renders the method potentially agnostic to the robot and touch sensor architectures while achieving a high degree of task adaptability.
- North America > Canada > Quebec > Estrie Region > Sherbrooke (0.14)
- North America > Canada > Ontario > Toronto (0.14)
- North America > Canada > Quebec > Montreal (0.14)
- (4 more...)
How Machine Learning is Revolutionising Video Game Development
Video games have come a long way since the early days of Pong and Space Invaders. Today, game developers are using cutting-edge technologies like machine learning to create more immersive and engaging gaming experiences than ever before. In this blog post, we'll explore how machine learning is transforming the video game industry and provide some practical advice for game developers looking to incorporate these technologies into their projects. One of the key applications of machine learning in video game development is player behaviour analysis. By using data analytics tools and machine learning algorithms, game developers can gain valuable insights into player behaviour, including how they play the game, what they like and dislike about it, and how they interact with other players.
Beauty maybe be skin deep, but AI finds revenue on the face's surface
A Taiwanese AI algorithm maker knows the value of the mind behind a face. The company says its software can perform virtual fashion try-ons and parse a consumer's personality with the same selfie. Perfect Corp. last week pushed a new AI and augmented reality makeup app and a fashion industry tie-in that could seed the market for high-end algorithmic social aspiration. In August, the company went down-market with a beard try-on product. None of that is to take away from Perfect's July decision to start selling the AI Personality Finder. It is a combination of facial-feature mapping and rudimentary psychological data that allegedly tells people not only how emotionally attractive they are, but also, what products to use to increase their visual likability.
- North America > United States > New York (0.06)
- Asia > Taiwan (0.06)